Subspace Clustering Reloaded: Sparse vs. Dense Representations

نویسندگان

  • Eva L. Dyer
  • Christoph Studer
  • Richard G. Baraniuk
چکیده

State-of-the-art methods for learning unions of subspaces from a collection of data leverage sparsity to form representations of each vector in the dataset with respect to the remaining vectors in the dataset. The resulting sparse representations can be used to form a subspace affinity matrix to cluster the data into their respective subspaces. While sparsity-driven methods for subspace clustering provide advantages over traditional nearest neighbor-based approaches, sparse representations often produce affinity matrices with weakly connected components that are difficult to cluster, even with state-of-the-art clustering methods. In this work, we propose a new algorithm that employs dense (least-squares) representations to extract the subspace affinity between vectors in the dataset. We demonstrate the advantages of the proposed dense subspace clustering algorithm over state-of-the-art sparsity-driven methods on real and synthetic data. I. SPARSE SUBSPACE CLUSTERING Unions of linear subspaces are a widely used signal model for representing collections of high-dimensional data, such as images of faces acquired under varying illumination conditions, motion trajectories from different objects, or local field potentials in brainmachine-interface applications [1], [2]. In order to use this signal model, the subspaces that the collection of data live upon must be learned from the data by performing subspace clustering—learning subspaces present in the data and clustering the data based upon the subspace membership of each vector. Subspace clustering is challenging due to the fact that subspace estimation and segmentation must be performed simultaneously. Sparse subspace clustering (SSC) has been shown to yield stateof-the-art performance on both synthetic and numerous real-world image datasets [1]. In SSC, a sparse representation of a signal in a collection of data is formed with respect to the remaining signals in the same dataset. The idea underlying this approach is that signals from the same subspace cluster will use one another in their sparse representations, thus revealing which points belong to the same subspace. SSC provides a powerful alternative to nearest neighbor (NN)-based approaches to subspace clustering and enables the derivation of guarantees that describe when convex methods [1], [3] or greedy algorithms [2] yield representations that only contain points from the same subspace. Formally, SSC computes sparse representations ci ∈ R for each point yi ∈ R, i = 1, . . . , d, via the following optimization problem: ci = argminc̃∈Rd ‖c̃‖1 subject to yi = ∑ j : j 6=i yj c̃j , (1) which are used to construct an affinity matrix C = [ c1· · · cd ]. Spectral clustering is then performed on the graph Laplacian of W = |C|+ |C | to segment the data into subspace clusters [2]. II. DENSE SUBSPACE CLUSTERING While sparse representations result in affinity matrices that contain a small number of edges in the graph linking signals from different subspaces, recovering subspace clusters from the affinity matrices obtained via SSC is challenging due to the fact that sparse representations often produce weakly connected components between signals in the same subspace. To circumvent this issue, we propose a novel method for subspace clustering that is based on forming dense leastsquares representations from the data. This method, which we refer 1We replace the `1 norm in (1) with the `2 norm resulting in (typically dense) least-squares representations of the data. DSC NN SSC

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Image Classification via Sparse Representation and Subspace Alignment

Image representation is a crucial problem in image processing where there exist many low-level representations of image, i.e., SIFT, HOG and so on. But there is a missing link across low-level and high-level semantic representations. In fact, traditional machine learning approaches, e.g., non-negative matrix factorization, sparse representation and principle component analysis are employed to d...

متن کامل

Multi-Task Learning for Subspace Segmentation

Subspace segmentation is the process of clustering a set of data points that are assumed to lie on the union of multiple linear or affine subspaces, and is increasingly being recognized as a fundamental tool for data analysis in high dimensional settings. Arguably one of the most successful approaches is based on the observation that the sparsest representation of a given point with respect to ...

متن کامل

Discriminative Transformation Learning for Fuzzy Sparse Subspace Clustering

This paper develops a novel iterative framework for subspace clustering (SC) in a learned discriminative feature domain. This framework consists of two modules of fuzzy sparse SC and discriminative transformation learning. In the first module, fuzzy latent labels containing discriminative information and latent representations capturing the subspace structure will be simultaneously evaluated in...

متن کامل

Sparse Additive Subspace Clustering

In this paper, we introduce and investigate a sparse additive model for subspace clustering problems. Our approach, named SASC (Sparse Additive Subspace Clustering), is essentially a functional extension of the Sparse Subspace Clustering (SSC) of Elhamifar & Vidal [7] to the additive nonparametric setting. To make our model computationally tractable, we express SASC in terms of a finite set of ...

متن کامل

Learning Stable Multilevel Dictionaries for Sparse Representation of Images

Dictionaries adapted to the data provide superior performance when compared to predefined dictionaries in applications involving sparse representations. Algorithmic stability and generalization are desirable characteristics for dictionary learning algorithms that aim to build global dictionaries which can efficiently model any test data similar to the training samples. In this paper, we propose...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2013